Goto

Collaborating Authors

 population count


PD-SEG: Population Disaggregation Using Deep Segmentation Networks For Improved Built Settlement Mask

Rahman, Muhammad Abdul, Waseem, Muhammad Ahmad, Khalid, Zubair, Tahir, Muhammad, Uppal, Momin

arXiv.org Artificial Intelligence

Any policy-level decision-making procedure and academic research involving the optimum use of resources for development and planning initiatives depends on accurate population density statistics. The current cutting-edge datasets offered by WorldPop and Meta do not succeed in achieving this aim for developing nations like Pakistan; the inputs to their algorithms provide flawed estimates that fail to capture the spatial and land-use dynamics. In order to precisely estimate population counts at a resolution of 30 meters by 30 meters, we use an accurate built settlement mask obtained using deep segmentation networks and satellite imagery. The Points of Interest (POI) data is also used to exclude non-residential areas.


So2Sat POP -- A Curated Benchmark Data Set for Population Estimation from Space on a Continental Scale

Doda, Sugandha, Wang, Yuanyuan, Kahl, Matthias, Hoffmann, Eike Jens, Ouan, Kim, Taubenböck, Hannes, Zhu, Xiao Xiang

arXiv.org Artificial Intelligence

Obtaining a dynamic population distribution is key to many decision-making processes such as urban planning, disaster management and most importantly helping the government to better allocate socio-technical supply. For the aspiration of these objectives, good population data is essential. The traditional method of collecting population data through the census is expensive and tedious. In recent years, statistical and machine learning methods have been developed to estimate population distribution. Most of the methods use data sets that are either developed on a small scale or not publicly available yet. Thus, the development and evaluation of new methods become challenging. We fill this gap by providing a comprehensive data set for population estimation in 98 European cities. The data set comprises a digital elevation model, local climate zone, land use proportions, nighttime lights in combination with multi-spectral Sentinel-2 imagery, and data from the Open Street Map initiative. We anticipate that it would be a valuable addition to the research community for the development of sophisticated approaches in the field of population estimation.


Fine-grained Population Mapping from Coarse Census Counts and Open Geodata

Metzger, Nando, Vargas-Muñoz, John E., Daudt, Rodrigo C., Kellenberger, Benjamin, Whelan, Thao Ton-That, Ofli, Ferda, Imran, Muhammad, Schindler, Konrad, Tuia, Devis

arXiv.org Artificial Intelligence

Fine-grained population maps are needed in several domains, like urban planning, environmental monitoring, public health, and humanitarian operations. Unfortunately, in many countries only aggregate census counts over large spatial units are collected, moreover, these are not always up-to-date. We present POMELO, a deep learning model that employs coarse census counts and open geodata to estimate fine-grained population maps with 100m ground sampling distance. Moreover, the model can also estimate population numbers when no census counts at all are available, by generalizing across countries. In a series of experiments for several countries in sub-Saharan Africa, the maps produced with POMELOare in good agreement with the most detailed available reference counts: disaggregation of coarse census counts reaches R2 values of 85-89%; unconstrained prediction in the absence of any counts reaches 48-69%.


Conversational BI: Text to SQL

#artificialintelligence

"the future of BI is Conversational" -- this is what Gartner and other analysts have been telling us for the last few years. Let's focus on structured data, relational data to be more precise. This forms the underlying storage format for most of the Business Intelligence (BI) world, irrespective of whether you are querying the database interactively or building a report in Tableau, Power BI, Qlik Sense, etc. The predominant language to interact with such storage platforms is SQL. We have already seen some products in this space, e.g., Power BI Q&A, Salesforce Photon.


The Census Is Broken. Can AI Fix It?

#artificialintelligence

Getting a census count wrong can cost communities big. A March 10 report from the US Census Bureau showed an overcount of white and Asian people and an undercount of people who identify as Black, Hispanic or Latino, or multiracial in 2020, a failure that has led to renewed calls to modernize the census. Progress reaching historically undercounted groups has been slow, and the stakes are high. The once-a-decade endeavor informs the distribution of federal tax dollars and apportions members of the House of Representatives for each state, potentially redrawing the political map. According to emails obtained through a records request, Trump administration officials interfered in the population count to produce outcomes beneficial to Republicans, but problems with the census go back much further.


The Census Is Broken. Can AI Fix It?

WIRED

Getting a census count wrong can cost communities big. A March 10 report from the US Census Bureau showed an overcount of white and Asian people and an undercount of people who identify as Black, Hispanic or Latino, or multiracial in 2020, a failure that has led to renewed calls to modernize the census. Progress reaching historically undercounted groups has been slow, and the stakes are high. The once-a-decade endeavor informs the distribution of federal tax dollars and apportions members of the House of Representatives for each state, potentially redrawing the political map. According to emails obtained through a records request, Trump administration officials interfered in the population count to produce outcomes beneficial to Republicans, but problems with the census go back much further.


The New Features of Python 3.10

#artificialintelligence

New features have been added that could be helpful while debugging the code, more information with pointing out the wrong syntax has been provided than just providing you with the "Syntax Error". For someone just starting in Python, Python 3.10 will come in handy when figuring out errors rather than using StackOverflow. As the name suggests, structural pattern matching is used in the form of match statements and case statements. The patterns could be sequences, mappings, primitive data types, and class instances. Finally, Python has a switch statement which is much more powerful than the switch statement.


How Data Visualization Can Help Correlate Diabetes And Income

#artificialintelligence

Is diabetes linked to wealth? Is it the new rich man's disease? Would work from home contribute to the rise in people with diabetes given the sedentary lifestyle amidst lockdowns? This analysis aims to dispel some common myths in the understanding distribution of diabetes among the common populace. It takes a partly economic and partly demographic approach towards finding important geographical centers for diabetes in terms of the total count, racial profile and income status, and its impact on diabetes count by counties in the USA. To analyze the relationship between different variables (income, racial profile by county) to diabetes in the USA and identify factors that most contribute to diabetes.


An aggregate learning approach for interpretable semi-supervised population prediction and disaggregation using ancillary data

Derval, Guillaume, Docquier, Frédéric, Schaus, Pierre

arXiv.org Machine Learning

Census data provide detailed information about population characteristics at a coarse resolution. Nevertheless, fine-grained, high-resolution mappings of population counts are increasingly needed to characterize population dynamics and to assess the consequences of climate shocks, natural disasters, investments in infrastructure, development policies, etc. Dissagregating these census is a complex machine learning, and multiple solutions have been proposed in past research. We propose in this paper to view the problem in the context of the aggregate learning paradigm, where the output value for all training points is not known, but where it is only known for aggregates of the points (i.e. in this context, for regions of pixels where a census is available). We demonstrate with a very simple and interpretable model that this method is on par, and even outperforms on some metrics, the state-of-the-art, despite its simplicity.